Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 12 de 12
Filtrar
1.
Sensors (Basel) ; 24(5)2024 Feb 28.
Artigo em Inglês | MEDLINE | ID: mdl-38475092

RESUMO

COVID-19 analysis from medical imaging is an important task that has been intensively studied in the last years due to the spread of the COVID-19 pandemic. In fact, medical imaging has often been used as a complementary or main tool to recognize the infected persons. On the other hand, medical imaging has the ability to provide more details about COVID-19 infection, including its severity and spread, which makes it possible to evaluate the infection and follow-up the patient's state. CT scans are the most informative tool for COVID-19 infection, where the evaluation of COVID-19 infection is usually performed through infection segmentation. However, segmentation is a tedious task that requires much effort and time from expert radiologists. To deal with this limitation, an efficient framework for estimating COVID-19 infection as a regression task is proposed. The goal of the Per-COVID-19 challenge is to test the efficiency of modern deep learning methods on COVID-19 infection percentage estimation (CIPE) from CT scans. Participants had to develop an efficient deep learning approach that can learn from noisy data. In addition, participants had to cope with many challenges, including those related to COVID-19 infection complexity and crossdataset scenarios. This paper provides an overview of the COVID-19 infection percentage estimation challenge (Per-COVID-19) held at MIA-COVID-2022. Details of the competition data, challenges, and evaluation metrics are presented. The best performing approaches and their results are described and discussed.


Assuntos
COVID-19 , Pandemias , Humanos , Benchmarking , Cintilografia , Tomografia Computadorizada por Raios X
2.
Front Neuroinform ; 17: 852105, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-36970658

RESUMO

Objective: In this study, we investigate whether a Convolutional Neural Network (CNN) can generate informative parametric maps from the pre-processed CT perfusion data in patients with acute ischemic stroke in a clinical setting. Methods: The CNN training was performed on a subset of 100 pre-processed perfusion CT dataset, while 15 samples were kept for testing. All the data used for the training/testing of the network and for generating ground truth (GT) maps, using a state-of-the-art deconvolution algorithm, were previously pre-processed using a pipeline for motion correction and filtering. Threefold cross validation had been used to estimate the performance of the model on unseen data, reporting Mean Squared Error (MSE). Maps accuracy had been checked through manual segmentation of infarct core and total hypo-perfused regions on both CNN-derived and GT maps. Concordance among segmented lesions was assessed using the Dice Similarity Coefficient (DSC). Correlation and agreement among different perfusion analysis methods were evaluated using mean absolute volume differences, Pearson correlation coefficients, Bland-Altman analysis, and coefficient of repeatability across lesion volumes. Results: The MSE was very low for two out of three maps, and low in the remaining map, showing good generalizability. Mean Dice scores from two different raters and the GT maps ranged from 0.80 to 0.87. Inter-rater concordance was high, and a strong correlation was found between lesion volumes of CNN maps and GT maps (0.99, 0.98, respectively). Conclusion: The agreement between our CNN-based perfusion maps and the state-of-the-art deconvolution-algorithm perfusion analysis maps, highlights the potential of machine learning methods applied to perfusion analysis. CNN approaches can reduce the volume of data required by deconvolution algorithms to estimate the ischemic core, and thus might allow the development of novel perfusion protocols with lower radiation dose deployed to the patient.

3.
Neural Netw ; 146: 230-237, 2022 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-34906759

RESUMO

LOBSTER (LOss-Based SensiTivity rEgulaRization) is a method for training neural networks having a sparse topology. Let the sensitivity of a network parameter be the variation of the loss function with respect to the variation of the parameter. Parameters with low sensitivity, i.e. having little impact on the loss when perturbed, are shrunk and then pruned to sparsify the network. Our method allows to train a network from scratch, i.e. without preliminary learning or rewinding. Experiments on multiple architectures and datasets show competitive compression ratios with minimal computational overhead.


Assuntos
Compressão de Dados , Redes Neurais de Computação
4.
IEEE Trans Neural Netw Learn Syst ; 33(12): 7237-7250, 2022 12.
Artigo em Inglês | MEDLINE | ID: mdl-34129503

RESUMO

Deep neural networks include millions of learnable parameters, making their deployment over resource-constrained devices problematic. Sensitivity-based regularization of neurons (SeReNe) is a method for learning sparse topologies with a structure, exploiting neural sensitivity as a regularizer. We define the sensitivity of a neuron as the variation of the network output with respect to the variation of the activity of the neuron. The lower the sensitivity of a neuron, the less the network output is perturbed if the neuron output changes. By including the neuron sensitivity in the cost function as a regularization term, we are able to prune neurons with low sensitivity. As entire neurons are pruned rather than single parameters, practical network footprint reduction becomes possible. Our experimental results on multiple network architectures and datasets yield competitive compression ratios with respect to state-of-the-art references.


Assuntos
Compressão de Dados , Redes Neurais de Computação , Compressão de Dados/métodos , Algoritmos , Neurônios
5.
Artigo em Inglês | MEDLINE | ID: mdl-32971995

RESUMO

The possibility to use widespread and simple chest X-ray (CXR) imaging for early screening of COVID-19 patients is attracting much interest from both the clinical and the AI community. In this study we provide insights and also raise warnings on what is reasonable to expect by applying deep learning to COVID classification of CXR images. We provide a methodological guide and critical reading of an extensive set of statistical results that can be obtained using currently available datasets. In particular, we take the challenge posed by current small size COVID data and show how significant can be the bias introduced by transfer-learning using larger public non-COVID CXR datasets. We also contribute by providing results on a medium size COVID CXR dataset, just collected by one of the major emergency hospitals in Northern Italy during the peak of the COVID pandemic. These novel data allow us to contribute to validate the generalization capacity of preliminary results circulating in the scientific community. Our conclusions shed some light into the possibility to effectively discriminate COVID using CXR.


Assuntos
Infecções por Coronavirus/diagnóstico por imagem , Aprendizado Profundo , Pneumonia Viral/diagnóstico por imagem , Radiografia Torácica , COVID-19 , Infecções por Coronavirus/epidemiologia , Conjuntos de Dados como Assunto , Humanos , Itália/epidemiologia , Pandemias , Pneumonia Viral/epidemiologia
6.
IEEE Trans Image Process ; 24(1): 205-19, 2015 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-25438310

RESUMO

The future of novel 3D display technologies largely depends on the design of efficient techniques for 3D video representation and coding. Recently, multiple view plus depth video formats have attracted many research efforts since they enable intermediate view estimation and permit to efficiently represent and compress 3D video sequences. In this paper, we present spatiotemporal occlusion compensation with panorama view (STOP), a novel 3D video coding technique based on the creation of a panorama view and occlusion coding in terms of spatiotemporal offsets. The panorama picture represents the most of the visual information acquired from multiple views using a single virtual view, characterized by a larger field of view. Encoding the panorama video with state-of-the-art HECV and representing occlusions with simple spatiotemporal ancillary information STOP achieves high-compression ratio and good visual quality with competitive results with respect to competing techniques. Moreover, STOP enables free viewpoint 3D TV applications whilst allowing legacy display to get a bidimensional service using a standard video codec and simple cropping operations.


Assuntos
Imageamento Tridimensional/métodos , Televisão , Algoritmos , Humanos , Gravação em Vídeo
7.
IEEE Trans Image Process ; 19(6): 1491-503, 2010 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-20215084

RESUMO

Digital fountain codes have emerged as a low-complexity alternative to Reed-Solomon codes for erasure correction. The applications of these codes are relevant especially in the field of wireless video, where low encoding and decoding complexity is crucial. In this paper, we introduce a new class of digital fountain codes based on a sliding-window approach applied to Raptor codes. These codes have several properties useful for video applications, and provide better performance than classical digital fountains. Then, we propose an application of sliding-window Raptor codes to wireless video broadcasting using scalable video coding. The rates of the base and enhancement layers, as well as the number of coded packets generated for each layer, are optimized so as to yield the best possible expected quality at the receiver side, and providing unequal loss protection to the different layers according to their importance. The proposed system has been validated in a UMTS broadcast scenario, showing that it improves the end-to-end quality, and is robust towards fluctuations in the packet loss rate.


Assuntos
Algoritmos , Artefatos , Compressão de Dados/métodos , Aumento da Imagem/métodos , Interpretação de Imagem Assistida por Computador/métodos , Processamento de Sinais Assistido por Computador , Telecomunicações , Reconhecimento Automatizado de Padrão/métodos , Reprodutibilidade dos Testes , Sensibilidade e Especificidade
8.
IEEE Trans Image Process ; 16(6): 1557-67, 2007 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-17547134

RESUMO

In this paper, an innovative joint-source channel coding scheme is presented. The proposed approach enables iterative soft decoding of arithmetic codes by means of a soft-in soft- out decoder based on suboptimal search and pruning of a binary tree. An error-resilient arithmetic coder with a forbidden symbol is used in order to improve the performance of the joint source/channel scheme. The performance in the case of transmission across the AWGN channel is evaluated in terms of word error probability and compared to a traditional separated approach. The interleaver gain, the convergence property of the system, and the optimal source/channel rate allocation are investigated. Finally, the practical relevance of the proposed joint decoding approach is demonstrated within the JPEG 2000 coding standard. In particular, an iterative channel and JPEG 2000 decoder is designed and tested in the case of image transmission across the AWGN channel.


Assuntos
Algoritmos , Redes de Comunicação de Computadores , Gráficos por Computador , Compressão de Dados/métodos , Aumento da Imagem/métodos , Interpretação de Imagem Assistida por Computador/métodos , Processamento de Sinais Assistido por Computador , Computação Matemática , Análise Numérica Assistida por Computador
9.
IEEE Trans Image Process ; 16(3): 673-83, 2007 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-17357728

RESUMO

In this paper, a novel multiple description coding technique is proposed, based on optimal Lagrangian rate allocation. The method assumes the coded data consists of independently coded blocks. Initially, all the blocks are coded at two different rates. Then blocks are split into two subsets with similar rate distortion characteristics; two balanced descriptions are generated by combining code blocks belonging to the two subsets encoded at opposite rates. A theoretical analysis of the approach is carried out, and the optimal rate distortion conditions are worked out. The method is successfully applied to the JPEG 2000 standard and simulation results show a noticeable performance improvement with respect to state-of-the art algorithms. The proposed technique enables easy tuning of the required coding redundancy. Moreover, the generated streams are fully compatible with Part 1 of the standard.


Assuntos
Algoritmos , Inteligência Artificial , Gráficos por Computador , Aumento da Imagem/métodos , Interpretação de Imagem Assistida por Computador/métodos , Armazenamento e Recuperação da Informação/métodos , Processamento de Sinais Assistido por Computador
10.
IEEE Trans Image Process ; 15(4): 807-18, 2006 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-16579370

RESUMO

JPEG 2000 is the novel ISO standard for image and video coding. Besides its improved coding efficiency, it also provides a few error resilience tools in order to limit the effect of errors in the codestream, which can occur when the compressed image or video data are transmitted over an error-prone channel, as typically occurs in wireless communication scenarios. However, for very harsh channels, these tools often do not provide an adequate degree of error protection. In this paper, we propose a novel error-resilience tool for JPEG 2000, based on the concept of ternary arithmetic coders employing a forbidden symbol. Such coders introduce a controlled degree of redundancy during the encoding process, which can be exploited at the decoder side in order to detect and correct errors. We propose a maximum likelihood and a maximum a posteriori context-based decoder, specifically tailored to the JPEG 2000 arithmetic coder, which are able to carry out both hard and soft decoding of a corrupted code-stream. The proposed decoder extends the JPEG 2000 capabilities in error-prone scenarios, without violating the standard syntax. Extensive simulations on video sequences show that the proposed decoders largely outperform the standard in terms of PSNR and visual quality.


Assuntos
Redes de Comunicação de Computadores , Gráficos por Computador , Compressão de Dados/métodos , Aumento da Imagem/métodos , Interpretação de Imagem Assistida por Computador/métodos , Processamento de Sinais Assistido por Computador , Gravação em Vídeo/métodos , Algoritmos , Compressão de Dados/normas , Aumento da Imagem/normas , Interpretação de Imagem Assistida por Computador/normas , Fotografação/métodos , Fotografação/normas , Viés de Seleção , Sensibilidade e Especificidade , Gravação em Vídeo/normas
11.
IEEE Trans Image Process ; 13(6): 751-7, 2004 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-15648866

RESUMO

We present hybrid loss protection as a new channel coding and packetization scheme for image transmission over nonprioritized lossy packet networks. The scheme employs an interleaver-based structure, and attempts to maximize the expected peak signal-to-noise ratio (PSNR) at the receiver given the constraint that the probability of failure, i.e., the probability that the PSNR of the decoded image is below a given threshold, is upper-bounded by a user-defined value. A new code-allocation algorithm is proposed, which employs Gilbert-Elliot modeling of the network statistics. Experimental results are provided in the case of transmission of images encoded by SPIHT and JPEG 2000 over a wireline, as well as a wireless UMTS-based Internet connection.


Assuntos
Algoritmos , Redes de Comunicação de Computadores , Aumento da Imagem/métodos , Interpretação de Imagem Assistida por Computador/métodos , Armazenamento e Recuperação da Informação/métodos , Garantia da Qualidade dos Cuidados de Saúde/métodos , Controle de Qualidade , Reprodutibilidade dos Testes , Sensibilidade e Especificidade
12.
IEEE Trans Image Process ; 11(6): 596-604, 2002.
Artigo em Inglês | MEDLINE | ID: mdl-18244658

RESUMO

This paper deals with the design and implementation of an image transform coding algorithm based on the integer wavelet transform (IWT). First of all, criteria are proposed for the selection of optimal factorizations of the wavelet filter polyphase matrix to be employed within the lifting scheme. The obtained results lead to the IWT implementations with very satisfactory lossless and lossy compression performance. Then, the effects of finite precision representation of the lifting coefficients on the compression performance are analyzed, showing that, in most cases, a very small number of bits can be employed for the mantissa keeping the performance degradation very limited. Stemming from these results, a VLSI architecture is proposed for the IWT implementation, capable of achieving very high frame rates with moderate gate complexity.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...